15 research outputs found

    Data-efficient deep representation learning

    Get PDF
    Current deep learning methods succeed in many data-intensive applications, but they are still not able to produce robust performance due to the lack of training samples. To investigate how to improve the performance of deep learning paradigms when training samples are limited, data-efficient deep representation learning (DDRL) is proposed in this study. DDRL as a sub area of representation learning mainly addresses the following problem: How can the performance of a deep learning method be maintained when the number of training samples is significantly reduced? This is vital for many applications where collecting data is highly costly, such as medical image analysis. Incorporating a certain kind of prior knowledge into the learning paradigm is key to achieving data efficiency. Deep learning as a sub-area of machine learning can be divided into three parts (locations) in its learning process, namely Data, Optimisation and Model. Integrating prior knowledge into these three locations is expected to bring data efficiency into a learning paradigm, which can dramatically increase the model performance under the condition of limited training data. In this thesis, we aim to develop novel deep learning methods for achieving data-efficient training, each of which integrates a certain kind of prior knowledge into three different locations respectively. We make the following contributions. First, we propose an iterative solution based on deep learning for medical image segmentation tasks, where dynamical systems are integrated into the segmentation labels in order to improve both performance and data efficiency. The proposed method not only shows a superior performance and better data efficiency compared to the state-of-the-art methods, but also has better interpretability and rotational invariance which are desired for medical imagining applications. Second, we propose a novel training framework which adaptively selects more informative samples for training during the optimization process. The adaptive selection or sampling is performed based on a hardness-aware strategy in the latent space constructed by a generative model. We show that the proposed framework outperforms a random sampling method, which demonstrates effectiveness of the proposed framework. Thirdly, we propose a deep neural network model which produces the segmentation maps in a coarse-to-fine manner. The proposed architecture is a sequence of computational blocks containing a number of convolutional layers in which each block provides its successive block with a coarser segmentation map as a reference. Such mechanisms enable us to train the network with limited training samples and produce more interpretable results.Open Acces

    VertXNet: Automatic Segmentation and Identification of Lumbar and Cervical Vertebrae from Spinal X-ray Images

    Full text link
    Manual annotation of vertebrae on spinal X-ray imaging is costly and time-consuming due to bone shape complexity and image quality variations. In this study, we address this challenge by proposing an ensemble method called VertXNet, to automatically segment and label vertebrae in X-ray spinal images. VertXNet combines two state-of-the-art segmentation models, namely U-Net and Mask R-CNN to improve vertebrae segmentation. A main feature of VertXNet is to also infer vertebrae labels thanks to its Mask R-CNN component (trained to detect 'reference' vertebrae) on a given spinal X-ray image. VertXNet was evaluated on an in-house dataset of lateral cervical and lumbar X-ray imaging for ankylosing spondylitis (AS) patients. Our results show that VertXNet can accurately label spinal X-rays (mean Dice of 0.9). It can be used to circumvent the lack of annotated vertebrae without requiring human expert review. This step is crucial to investigate clinical associations by solving the lack of segmentation, a common bottleneck for most computational imaging projects

    Additional file 1 of Supporting systematic reviews using LDA-based document representations

    No full text
    Supplementary figures. Specific results achieved for the other corpora. (DOCX 368 kb

    Supporting systematic reviews using LDA-based document representations

    No full text
    Abstract Background Identifying relevant studies for inclusion in a systematic review (i.e. screening) is a complex, laborious and expensive task. Recently, a number of studies has shown that the use of machine learning and text mining methods to automatically identify relevant studies has the potential to drastically decrease the workload involved in the screening phase. The vast majority of these machine learning methods exploit the same underlying principle, i.e. a study is modelled as a bag-of-words (BOW). Methods We explore the use of topic modelling methods to derive a more informative representation of studies. We apply Latent Dirichlet allocation (LDA), an unsupervised topic modelling approach, to automatically identify topics in a collection of studies. We then represent each study as a distribution of LDA topics. Additionally, we enrich topics derived using LDA with multi-word terms identified by using an automatic term recognition (ATR) tool. For evaluation purposes, we carry out automatic identification of relevant studies using support vector machine (SVM)-based classifiers that employ both our novel topic-based representation and the BOW representation. Results Our results show that the SVM classifier is able to identify a greater number of relevant studies when using the LDA representation than the BOW representation. These observations hold for two systematic reviews of the clinical domain and three reviews of the social science domain. Conclusions A topic-based feature representation of documents outperforms the BOW representation when applied to the task of automatic citation screening. The proposed term-enriched topics are more informative and less ambiguous to systematic reviewers

    The value of the ACEF II score in Chinese patients with elective and non-elective cardiac surgery

    No full text
    Abstract Objective To evaluate the value of the ACEF II score in predicting postoperative hospital death and acute kidney injury requiring dialysis (AKI-D) in Chinese patients. Methods This retrospective study included adult patients who underwent cardiopulmonary bypass open heart surgery between January 2010 and December 2015 at Guangdong Provincial People’s Hospital. ACEF II was evaluated to predict in-hospital death and AKI-D using the Hosmer–Lemeshow goodness of fit test for calibration and area under the receiver operating characteristic (ROC) curve for discrimination in non-elective and elective cardiac surgery. Results A total of 9748 patients were included. Among them, 1080 underwent non-elective surgery, and 8615 underwent elective surgery. Mortality was 1.8% (177/9748). In elective surgery, the area under the ROC (AUC) of the ACEF II score was 0.704 (95% CI: 0.648–0.759), similar to the ACEF score of 0.709 (95% CI: 0.654–0.763). In non-elective surgery, the AUC of the ACEF II score was 0.725 (95% CI: 0.663–0.787), higher than the ACEF score (AUC = 0.625, 95% CI: 0.553–0.697). The incidence of AKI-D was 3.5% (345/9748). The AUC of the ACEF II score was 0.718 (95% CI: 0.687–0.749), higher than the ACEF score (AUC = 0.626, 95% CI: 0.594–0.658). Conclusion ACEF and ACEF II have poor discrimination ability in predicting AKI-D in non-elective surgery. The ACEF II and ACEF scores have the same ability to predict in-hospital death in elective cardiac surgery, and the ACEF II score is better in non-elective surgery. The ACEF II score can be used to assess the risk of AKI-D in elective surgery in Chinese adults
    corecore